haptic rendering
Active Learning of Fractional-Order Viscoelastic Model Parameters for Realistic Haptic Rendering
Tolasa, Harun, Gemalmaz, Gorkem, Patoglu, Volkan
Fractional-order models provide an effective means of describing intrinsically time-dependent viscoelastic dynamics with few parameters, as these models can naturally capture memory effects. However, due to the unintuitive frequency-dependent coupling between the order of the fractional element and the other parameters, determining appropriate parameters for fractional-order models that yield high perceived realism remains a significant challenge. In this study, we propose a systematic means of determining the parameters of fractional-order viscoelastic models that optimizes the perceived realism of haptic rendering across general populations. First, we demonstrate that the parameters of fractional-order models can be effectively optimized through active learning, via qualitative feedback-based human-in-the-loop (HiL) optimizations, to ensure consistently high realism ratings for each individual. Second, we propose a rigorous method to combine HiL optimization results to form an aggregate perceptual map trained on the entire dataset and demonstrate the selection of population-level optimal parameters from this representation that are broadly perceived as realistic across general populations. Finally, we provide evidence of the effectiveness of the generalized fractional-order viscoelastic model parameters by characterizing their perceived realism through human-subject experiments. Overall, generalized fractional-order viscoelastic models established through the proposed HiL optimization and aggregation approach possess the potential to significantly improve the sim-to-real transition performance of medical training simulators. Index T erms--Viscoelastic materials, fractional-order standard linear solid model, haptic rendering, human-in-the-loop optimization, perceived realism, and medical training simulators.
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > New Jersey > Bergen County > Mahwah (0.04)
- North America > United States > Michigan > Washtenaw County > Ann Arbor (0.04)
- (3 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.88)
- Education (1.00)
- Health & Medicine > Therapeutic Area (0.93)
- Leisure & Entertainment > Games > Computer Games (0.34)
The Effect of Haptic Guidance during Robotic-assisted Motor Training is Modulated by Personality Traits
Garzás-Villar, Alberto, Boersma, Caspar, Derumigny, Alexis, Zgonnikov, Arkady, Marchal-Crespo, Laura
The provision of robotic assistance during motor training has proven to be effective in enhancing motor learning in some healthy trainee groups as well as patients. Personalizing such robotic assistance can help further improve motor (re)learning outcomes and cater better to the trainee's needs and desires. However, the development of personalized haptic assistance is hindered by the lack of understanding of the link between the trainee's personality and the effects of haptic guidance during human-robot interaction. To address this gap, we ran an experiment with 42 healthy participants who trained with a robotic device to control a virtual pendulum to hit incoming targets either with or without haptic guidance. We found that certain personal traits affected how users adapt and interact with the guidance during training. In particular, those participants with an 'Achiever gaming style' performed better and applied lower interaction forces to the robotic device than the average participant as the training progressed. Conversely, participants with the 'Free spirit game style' increased the interaction force in the course of training. We also found an interaction between some personal characteristics and haptic guidance. Specifically, participants with a higher 'Transformation of challenge' trait exhibited poorer performance during training while receiving haptic guidance compared to an average participant receiving haptic guidance. Furthermore, individuals with an external Locus of Control tended to increase their interaction force with the device, deviating from the pattern observed in an average participant under the same guidance. These findings suggest that individual characteristics may play a crucial role in the effectiveness of haptic guidance training strategies.
- Europe > Netherlands > South Holland > Delft (0.05)
- Europe > Netherlands > South Holland > Rotterdam (0.05)
- Europe > Switzerland (0.04)
- North America (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
Vis2Hap: Vision-based Haptic Rendering by Cross-modal Generation
Cao, Guanqun, Jiang, Jiaqi, Mao, Ningtao, Bollegala, Danushka, Li, Min, Luo, Shan
To assist robots in teleoperation tasks, haptic rendering which allows human operators access a virtual touch feeling has been developed in recent years. Most previous haptic rendering methods strongly rely on data collected by tactile sensors. However, tactile data is not widely available for robots due to their limited reachable space and the restrictions of tactile sensors. To eliminate the need for tactile data, in this paper we propose a novel method named as Vis2Hap to generate haptic rendering from visual inputs that can be obtained from a distance without physical interaction. We take the surface texture of objects as key cues to be conveyed to the human operator. To this end, a generative model is designed to simulate the roughness and slipperiness of the object's surface. To embed haptic cues in Vis2Hap, we use height maps from tactile sensors and spectrograms from friction coefficients as the intermediate outputs of the generative model. Once Vis2Hap is trained, it can be used to generate height maps and spectrograms of new surface textures, from which a friction image can be obtained and displayed on a haptic display. The user study demonstrates that our proposed Vis2Hap method enables users to access a realistic haptic feeling similar to that of physical objects. The proposed vision-based haptic rendering has the potential to enhance human operators' perception of the remote environment and facilitate robotic manipulation.
- Europe > United Kingdom (0.14)
- Asia > China > Shaanxi Province > Xi'an (0.04)